Before testing, ensure your ZK_LIST property is set correctly at line 8 within distributed-louvain-modularity/louvain.py.


In [1]:
cd distributed-louvain-modularity/example


/home/vagrant/distributed-louvain-modularity/example

In [2]:
./run_example.sh


clear out old example dir..Deleted /tmp/louvain-giraph-example
loading example data to /tmp/louvain-giraph-examplerunning the full louvain pipeline

running full louvain on /tmp/louvain-giraph-example/input
saving output to: /tmp/louvain-giraph-example/output

Removing ouput dirDeleted /tmp/louvain-giraph-example/output

running giraph phase /tmp/louvain-giraph-example/input -> /tmp/louvain-giraph-example/output/giraph_1
rm: `/tmp/louvain-giraph-example/output/giraph_1': No such file or directory14/01/26 03:24:41 INFO utils.ConfigurationUtils: No edge input format specified. Ensure your InputFormat does not require one.
14/01/26 03:24:41 INFO utils.ConfigurationUtils: Setting custom argument [giraph.vertex.input.dir] to [/tmp/louvain-giraph-example/input] in GiraphConfiguration
14/01/26 03:24:41 INFO utils.ConfigurationUtils: Setting custom argument [ mapreduce.task.timeout] to [10800000] in GiraphConfiguration
14/01/26 03:24:41 INFO utils.ConfigurationUtils: Setting custom argument [actual.Q.aggregators] to [1] in GiraphConfiguration
14/01/26 03:24:41 INFO utils.ConfigurationUtils: Setting custom argument [minimum.progress] to [2000] in GiraphConfiguration
14/01/26 03:24:41 INFO utils.ConfigurationUtils: Setting custom argument [progress.tries] to [1] in GiraphConfiguration14/01/26 03:24:41 INFO job.GiraphJob: run: Since checkpointing is disabled (default), do not allow any task retries (setting mapred.map.max.attempts = 0, old value = 4)14/01/26 03:24:41 WARN mapred.JobClient: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same.14/01/26 03:24:43 INFO mapred.JobClient: Running job: job_201401260304_000214/01/26 03:24:44 INFO mapred.JobClient:  map 0% reduce 0%14/01/26 03:24:59 INFO mapred.JobClient:  map 50% reduce 0%14/01/26 03:25:00 INFO mapred.JobClient:  map 100% reduce 0%14/01/26 03:25:02 INFO mapred.JobClient: Job complete: job_201401260304_0002
14/01/26 03:25:02 INFO mapred.JobClient: Counters: 34
14/01/26 03:25:02 INFO mapred.JobClient:   File System Counters
14/01/26 03:25:02 INFO mapred.JobClient:     FILE: Number of bytes read=0
14/01/26 03:25:02 INFO mapred.JobClient:     FILE: Number of bytes written=410326
14/01/26 03:25:02 INFO mapred.JobClient:     FILE: Number of read operations=0
14/01/26 03:25:02 INFO mapred.JobClient:     FILE: Number of large read operations=0
14/01/26 03:25:02 INFO mapred.JobClient:     FILE: Number of write operations=0
14/01/26 03:25:02 INFO mapred.JobClient:     HDFS: Number of bytes read=432
14/01/26 03:25:02 INFO mapred.JobClient:     HDFS: Number of bytes written=134
14/01/26 03:25:02 INFO mapred.JobClient:     HDFS: Number of read operations=10
14/01/26 03:25:02 INFO mapred.JobClient:     HDFS: Number of large read operations=0
14/01/26 03:25:02 INFO mapred.JobClient:     HDFS: Number of write operations=4
14/01/26 03:25:02 INFO mapred.JobClient:   Job Counters 
14/01/26 03:25:02 INFO mapred.JobClient:     Launched map tasks=2
14/01/26 03:25:02 INFO mapred.JobClient:     Total time spent by all maps in occupied slots (ms)=26203
14/01/26 03:25:02 INFO mapred.JobClient:     Total time spent by all reduces in occupied slots (ms)=0
14/01/26 03:25:02 INFO mapred.JobClient:     Total time spent by all maps waiting after reserving slots (ms)=0
14/01/26 03:25:02 INFO mapred.JobClient:     Total time spent by all reduces waiting after reserving slots (ms)=014/01/26 03:25:02 INFO mapred.JobClient:   Map-Reduce Framework
14/01/26 03:25:02 INFO mapred.JobClient:     Map input records=2
14/01/26 03:25:02 INFO mapred.JobClient:     Map output records=0
14/01/26 03:25:02 INFO mapred.JobClient:     Input split bytes=88
14/01/26 03:25:02 INFO mapred.JobClient:     Spilled Records=0
14/01/26 03:25:02 INFO mapred.JobClient:     CPU time spent (ms)=2690
14/01/26 03:25:02 INFO mapred.JobClient:     Physical memory (bytes) snapshot=269602816
14/01/26 03:25:02 INFO mapred.JobClient:     Virtual memory (bytes) snapshot=3734609920
14/01/26 03:25:02 INFO mapred.JobClient:     Total committed heap usage (bytes)=253100032
14/01/26 03:25:02 INFO mapred.JobClient:   Giraph Stats
14/01/26 03:25:02 INFO mapred.JobClient:     Aggregate edges=11
14/01/26 03:25:02 INFO mapred.JobClient:     Aggregate finished vertices=0
14/01/26 03:25:02 INFO mapred.JobClient:     Aggregate vertices=9
14/01/26 03:25:02 INFO mapred.JobClient:     Current master task partition=0
14/01/26 03:25:02 INFO mapred.JobClient:     Current workers=1
14/01/26 03:25:02 INFO mapred.JobClient:     Last checkpointed superstep=0
14/01/26 03:25:02 INFO mapred.JobClient:     Sent messages=0
14/01/26 03:25:02 INFO mapred.JobClient:     Superstep=14
14/01/26 03:25:02 INFO mapred.JobClient:   Giraph Timers
14/01/26 03:25:02 INFO mapred.JobClient:     Setup (milliseconds)=1167
14/01/26 03:25:02 INFO mapred.JobClient:     Shutdown (milliseconds)=232
14/01/26 03:25:02 INFO mapred.JobClient:     Total (milliseconds)=2741giraph exit status: 0 time (sec): 27.4048581123Found 1 items
-rw-r--r--   1 vagrant supergroup         18 2014-01-26 03:24 /tmp/louvain-giraph-example/output/_COMPLETE
giraph times 27.4048581123
total giraph time: 27.4048581123
mapreduce times 
total mapreduce time 0
Total time: 27.4048581123

In [3]:
hadoop fs -get /tmp/louvain-giraph-example/output .

In [4]:
cd output


/home/vagrant/distributed-louvain-modularity/example/output

In [5]:
ls -la


total 20
drwxrwxr-x 3 vagrant vagrant 4096 Jan 26 03:25 ./
drwxrwxr-x 3 vagrant vagrant 4096 Jan 26 03:25 ../
-rwxr-xr-x 1 vagrant vagrant   18 Jan 26 03:25 _COMPLETE*
drwxrwxr-x 3 vagrant vagrant 4096 Jan 26 03:25 giraph_1/
-rwxr-xr-x 1 vagrant vagrant   18 Jan 26 03:25 _q_1*

In [6]:
cat _COMPLETE


0.4134948096885813

In [7]:
cd giraph_1/


/home/vagrant/distributed-louvain-modularity/example/output/giraph_1

In [8]:
ls -la


total 16
drwxrwxr-x 3 vagrant vagrant 4096 Jan 26 03:25 ./
drwxrwxr-x 3 vagrant vagrant 4096 Jan 26 03:25 ../
drwxrwxr-x 3 vagrant vagrant 4096 Jan 26 03:25 _logs/
-rwxr-xr-x 1 vagrant vagrant   98 Jan 26 03:25 part-m-00001*
-rwxr-xr-x 1 vagrant vagrant    0 Jan 26 03:25 _SUCCESS*

In [9]:
cat part-m-00001


4	3	0	3:4
2	3	0	3:4
9	8	0	8:3
5	3	0	3:4
6	8	0	3:1,8:3
3	3	0	3:4
7	8	0	8:3
1	3	0	3:4,8:1
8	8	0	8:3

In [10]:
cat ../../small.tsv


1	0	2:1,3:1,4:1,5:1,6:1
2	0	1:1,3:1,4:1,5:1
3	0	1:1,2:1,4:1,5:1
4	0	1:1,2:1,3:1,5:1
5	0	1:1,2:1,3:1,4:1
6	0	1:1,7:1,8:1,9:1
7	0	6:1,8:1,9:1
8	0	6:1,7:1,9:1
9	0	6:1,7:1,8:1

In [ ]: